Goto

Collaborating Authors

 deep matrix factorization


Implicit Regularization in Deep Matrix Factorization

Neural Information Processing Systems

Efforts to understand the generalization mystery in deep learning have led to the belief that gradient-based optimization induces a form of implicit regularization, a bias towards models of low complexity. We study the implicit regularization of gradient descent over deep linear neural networks for matrix completion and sensing, a model referred to as deep matrix factorization. Our first finding, supported by theory and experiments, is that adding depth to a matrix factorization enhances an implicit tendency towards low-rank solutions, oftentimes leading to more accurate recovery. Secondly, we present theoretical and empirical arguments questioning a nascent view by which implicit regularization in matrix factorization can be captured using simple mathematical norms. Our results point to the possibility that the language of standard regularizers may not be rich enough to fully encompass the implicit regularization brought forth by gradient-based optimization.


Computational Mirrors: Blind Inverse Light Transport by Deep Matrix Factorization

Neural Information Processing Systems

We recover a video of the motion taking place in a hidden scene by observing changes in indirect illumination in a nearby uncalibrated visible region. We solve this problem by factoring the observed video into a matrix product between the unknown hidden scene video and an unknown light transport matrix. This task is extremely ill-posed, as any non-negative factorization will satisfy the data. Inspired by recent work on the Deep Image Prior, we parameterize the factor matrices using randomly initialized convolutional neural networks trained in a one-off manner, and show that this results in decompositions that reflect the true motion in the hidden scene.


Sharpness of Minima in Deep Matrix Factorization: Exact Expressions

Kamber, Anil, Parhi, Rahul

arXiv.org Machine Learning

Understanding the geometry of the loss landscape near a minimum is key to explaining the implicit bias of gradient-based methods in non-convex optimization problems such as deep neural network training and deep matrix factorization. A central quantity to characterize this geometry is the maximum eigenvalue of the Hessian of the loss, which measures the sharpness of the landscape. Currently, its precise role has been obfuscated because no exact expressions for this sharpness measure were known in general settings. In this paper, we present the first exact expression for the maximum eigenvalue of the Hessian of the squared-error loss at any minimizer in general overparameterized deep matrix factorization (i.e., deep linear neural network training) problems, resolving an open question posed by Mulayoff & Michaeli (2020). This expression uncovers a fundamental property of the loss landscape of depth-2 matrix factorization problems: a minimum is flat if and only if it is spectral-norm balanced, which implies that flat minima are not necessarily Frobenius-norm balanced. Furthermore, to complement our theory, we empirically investigate an escape phenomenon observed during gradient-based training near a minimum that crucially relies on our exact expression of the sharpness. Decades of research in learning theory suggest limiting model complexity to prevent overfitting.


Implicit Regularization in Deep Matrix Factorization

Neural Information Processing Systems

A major hurdle in this study is that implicit regularization in deep learning seems to kick in only with certain types of data (not with random data for example), and we lack mathematical tools for reasoning about real-life data. Thus one needs a simple test-bed for the investigation, where data admits a crisp mathematical formulation.


Reviews: Implicit Regularization in Deep Matrix Factorization

Neural Information Processing Systems

This paper studies the implicit regularization of gradient descent over deep neural networks for deep matrix factorization models. The paper begins with a review of prior work regarding how running gradient descent on a shallow matrix factorization model, with small learning rate and initialization close to zero, tends to converge to solutions that minimize the nuclear norm [20] (Conjecture 1). This discussion is then extended to deep matrix factorization, where predictive performance improves with depth when the number of observed entries is small. Experimental results (Figure 2) which challenge Conjecture 1 are then presented, which indicate that implicit regularization in both shallow and deep matrix factorization converges to low-rank solutions, rather than minimizing nuclear norm, when few entries are observed. Finally, a theoretical and experimental analysis of the dynamics of gradient flow for deep matrix factorization is presented, which shows how singular values and singular vectors of the product matrix evolve during training, and how this leads to implicit regularization that induces low-rank solutions.


Reviews: Implicit Regularization in Deep Matrix Factorization

Neural Information Processing Systems

The reviewers were unanimous in liking the paper. They make multiple presentational suggestions that the authors should incorporate in their final version: some figures are hard to read, as well as additional pointed discussion in certain sections such as Sec 2.2 (which the authors agreed to in their rebuttal).


Reviews: Computational Mirrors: Blind Inverse Light Transport by Deep Matrix Factorization

Neural Information Processing Systems

In this paper, the authors study the problem of reconstructing a hidden scene from the observed videos. The proposed method seeks to invert tight transport matrix without a calibration step. The problem is challenging and ill posed. The author learn a low-dimensional basis from observed videos and use deep image prior models for generating hidden scene and coefficients of the light transport basis. Originality: The paper uses inverse light transport to recover a video of hidden scene without any calibration, which seems novel.


Reviews: Computational Mirrors: Blind Inverse Light Transport by Deep Matrix Factorization

Neural Information Processing Systems

This paper has tackled an extremely challenging problem. It provides a neat and bold idea towards solving it. While the work is far from complete, as agreed upon by many of the reviewers in a discussion, it provides a first-cut idea and attempt, and enough detail to potentially carry this proof-of-concept further in the future. I suggest the authors carefully address the reviewer's comments.


Deep Matrix Factorization with Adaptive Weights for Multi-View Clustering

Khalafaoui, Yasser, Matei, Basarab, Lovisetto, Martino, Grozavu, Nistor

arXiv.org Machine Learning

Recently, deep matrix factorization has been established as a powerful model for unsupervised tasks, achieving promising results, especially for multi-view clustering. However, existing methods often lack effective feature selection mechanisms and rely on empirical hyperparameter selection. To address these issues, we introduce a novel Deep Matrix Factorization with Adaptive Weights for Multi-View Clustering (DMFAW). Our method simultaneously incorporates feature selection and generates local partitions, enhancing clustering results. Notably, the features weights are controlled and adjusted by a parameter that is dynamically updated using Control Theory inspired mechanism, which not only improves the model's stability and adaptability to diverse datasets but also accelerates convergence. A late fusion approach is then proposed to align the weighted local partitions with the consensus partition. Finally, the optimization problem is solved via an alternating optimization algorithm with theoretically guaranteed convergence. Extensive experiments on benchmark datasets highlight that DMFAW outperforms state-of-the-art methods in terms of clustering performance.


Implicit Regularization in Deep Matrix Factorization

Neural Information Processing Systems

Efforts to understand the generalization mystery in deep learning have led to the belief that gradient-based optimization induces a form of implicit regularization, a bias towards models of low "complexity." We study the implicit regularization of gradient descent over deep linear neural networks for matrix completion and sensing, a model referred to as deep matrix factorization. Our first finding, supported by theory and experiments, is that adding depth to a matrix factorization enhances an implicit tendency towards low-rank solutions, oftentimes leading to more accurate recovery. Secondly, we present theoretical and empirical arguments questioning a nascent view by which implicit regularization in matrix factorization can be captured using simple mathematical norms. Our results point to the possibility that the language of standard regularizers may not be rich enough to fully encompass the implicit regularization brought forth by gradient-based optimization.